Variable selection with false discovery rate control in deep neural networks

نویسندگان

چکیده

Deep neural networks are famous for their high prediction accuracy, but they also known black-box nature and poor interpretability. We consider the problem of variable selection, that is, selecting input variables have significant predictive power on output, in deep networks. Most existing selection methods only applicable to shallow or computationally infeasible large datasets; moreover, lack a control quality selected variables. Here we propose backward elimination procedure called SurvNet, which is based new measure importance applies wide variety More importantly, SurvNet able estimate false discovery rate empirically. Further, adaptively determines how many eliminate at each step order maximize efficiency. The validity efficiency shown various simulated real datasets, its performance compared with other methods. Especially, systematic comparison knockoff-based shows although more rigorous data strong correlation, usually has higher power. Identifying salient features can be challenge authors developed works classification regression problems, one multiple output neurons,

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

False Discovery Rate Control With Groups.

In the context of large-scale multiple hypothesis testing, the hypotheses often possess certain group structures based on additional information such as Gene Ontology in gene expression data and phenotypes in genome-wide association studies. It is hence desirable to incorporate such information when dealing with multiplicity problems to increase statistical power. In this article, we demonstrat...

متن کامل

Private False Discovery Rate Control

We provide the first differentially private algorithms for controlling the false discovery rate (FDR) in multiple hypothesis testing, with essentially no loss in power under certain conditions. Our general approach is to adapt a well-known variant of the Benjamini-Hochberg procedure (BHq), making each step differentially private. This destroys the classical proof of FDR control. To prove FDR co...

متن کامل

Variable selection with neural networks

In this paper, we present 3 different neural network-based methods to perform variable selection. OCD Optimal Cell Damage is a pruning method, which evaluates the usefulness of a variable and prunes the least useful ones (it is related to the Optimal Brain Damage method of J_.e Cun et al.). Regularization theory proposes to constrain estimators by adding a term to the cost function used to trai...

متن کامل

Gene Selection using Multidimensional False Discovery Rate

This paper proposes our algorithm for gene selection in microarray data analysis comparing conditions with replicates. Based on background noise computation in replicate array, this algorithm uses the global False Discovery Rate based on ‘Between’ group and ‘Within’ group comparisons of replicates to select the set of differential expressed genes. This method uses two types of statistics that l...

متن کامل

Multiple Attribute Control Charts with False Discovery Rate Control

The statistical cumulative sum (CUSUM) chart is a powerful tool for monitoring the attribute quality variable in manufacturing industry. In this article, we studied the multiplicity problem caused by simultaneously monitoring more than one attribute quality variable. Multiple binomial and Poisson CUSUM charts incorporating a multiple hypothesis testing technique known as false discovery rate co...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Nature Machine Intelligence

سال: 2021

ISSN: ['2522-5839']

DOI: https://doi.org/10.1038/s42256-021-00308-z